Fast bootstrap methodology for regression model selection

نویسندگان

  • Amaury Lendasse
  • Geoffroy Simon
  • Vincent Wertz
  • Michel Verleysen
چکیده

Using resampling methods like cross-validation and bootstrap is a necessity in neural network design, for solving the problem of model structure selection. The bootstrap is a powerful method offering a low variance of the model generalization error estimate. Unfortunately, its computational load may be excessive when used to select among neural networks models of different structures or complexities. This paper presents the fast bootstrap (FB) methodology to select the best model structure; this methodology is applied here to regression tasks. The fast bootstrap assumes that the computationally expensive term estimated by the bootstrap, the optimism, is usually a smooth function (low-order polynomial) of the complexity parameter. Approximating the optimism term makes it possible to considerably reduce the necessary number of simulations. The FB methodology is illustrated on multi-layer perceptrons, radial-basis function networks and least-square support vector machines. r 2004 Published by Elsevier B.V.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Robust model selection using fast and robust bootstrap

Robust model selection procedures control the undue influence that outliers can have on the selection criteria by using both robust point estimators and a bounded loss function when measuring either the goodness-of-fit or the expected prediction error of each model. Furthermore, to avoid favoring over-fitting models, these two measures can be combined with a penalty term for the size of the mod...

متن کامل

Robust Model Selection in Generalized Linear Models

In this paper, we extend to generalized linear models (including logistic and other binary regression models, Poisson regression and gamma regression models) the robust model selection methodology developed by Müller and Welsh (2005) for linear regression models. As in Müller and Welsh (2005), we combine a robust penalized measure of fit to the sample with a robust measure of out of sample pred...

متن کامل

Model Selection, Estimation, and Bootstrap Smoothing

Classical statistical theory ignores model selection in assessing estimation accuracy. Here we consider bootstrap methods for computing standard errors and confidence intervals that take model selection into account. The methodology involves bootstrap smoothing, also known as bagging, to tame the erratic discontinuities of selection-based estimators. A projection theorem then provides standard ...

متن کامل

Fast approximation of the bootstrap for model selection

The bootstrap resampling method may be efficiently used to estimate the generalization error of a family of nonlinear regression models, as artificial neural networks. The main difficulty associated with the bootstrap in real-world applications is the high computation load. In this paper we propose a simple procedure based on empirical evidence, to considerably reduce the computation time neede...

متن کامل

Regularization parameter selection in indirect regression by residual based bootstrap

Residual-based analysis is generally considered a cornerstone of statistical methodology. For a special case of indirect regression, we investigate the residual-based empirical distribution function and provide a uniform expansion of this estimator, which is also shown to be asymptotically most precise. This investigation naturally leads to a completely data-driven technique for selecting a reg...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • Neurocomputing

دوره 64  شماره 

صفحات  -

تاریخ انتشار 2005